2026-04-10 19:11:38,697 [ 214304 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2026-04-10 19:11:38,697 [ 214304 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:79, check_args_and_update_paths) 2026-04-10 19:11:38,697 [ 214304 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:90, check_args_and_update_paths) 2026-04-10 19:11:38,697 [ 214304 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:92, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_vlduwg --privileged --dns-search='.' --memory=30709047296 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=5ccda723c1fc -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=d862517635bf -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=1 --color=no --durations=0 --report-log=parallel0_1.jsonl --report-log-exclude-logs-on-passed-tests test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup -vvv " altinityinfra/integration-tests-runner:226bfaf75ac1 '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache Test order randomisation NOT enabled. Enable with --random-order or --random-order-bucket= rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: timeout-2.3.1, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0, random-order-1.1.1 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [1 item] scheduling tests via LoadFileScheduling test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup [gw0] [100%] FAILED test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup =================================== FAILURES =================================== _________________________ test_shutdown_cancels_backup _________________________ [gw0] linux -- Python 3.10.12 /usr/bin/python3 def test_shutdown_cancels_backup(): > with NoTrashChecker() as no_trash_checker: test_backup_restore_on_cluster/test_cancel_backup.py:556: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = type = None, value = None, traceback = None def __exit__(self, type, value, traceback): list_of_znodes = set( node1.query( "SELECT name FROM system.zookeeper WHERE path = '/clickhouse/backups' " + "AND NOT (name == 'alive_tracker')" ).splitlines() ) new_znodes = list_of_znodes.difference(self.__previous_list_of_znodes) if new_znodes: print(f"Found nodes in ZooKeeper: {new_znodes}") for node in new_znodes: print( f"Nodes in '/clickhouse/backups/{node}':\n" + node1.query( f"SELECT name FROM system.zookeeper WHERE path = '/clickhouse/backups/{node}'" ) ) print( f"Nodes in '/clickhouse/backups/{node}/stage':\n" + node1.query( f"SELECT name FROM system.zookeeper WHERE path = '/clickhouse/backups/{node}/stage'" ) ) if self.check_zookeeper: assert new_znodes == set() list_of_backups = set( os.listdir(os.path.join(node1.cluster.instances_dir, "backups")) ) new_backups = list_of_backups.difference(self.__previous_list_of_backups) unfinished_backups = set( backup for backup in new_backups if not os.path.exists( os.path.join(node1.cluster.instances_dir, "backups", backup, ".backup") ) ) new_backups = set( backup for backup in new_backups if backup not in unfinished_backups ) if new_backups: print(f"Found new backups: {new_backups}") if unfinished_backups: print(f"Found unfinished backups: {unfinished_backups}") assert new_backups == set(self.expect_backups) assert unfinished_backups.difference(self.allow_unfinished_backups) == set() all_errors = set() start_time = time.strftime( "%Y-%m-%d %H:%M:%S", self.__start_time_for_collecting_errors ) for node in nodes: errors_query_result = node.query( "SELECT name FROM system.errors WHERE last_error_time >= toDateTime('" + start_time + "') " + "AND NOT ((name == 'KEEPER_EXCEPTION') AND (last_error_message LIKE '%Fault injection%')) " + "AND NOT (name == 'NO_ELEMENTS_IN_CONFIG')" ) errors = errors_query_result.splitlines() if errors: print(f"{get_node_name(node)}: Found errors: {errors}") print( node.query( "SELECT name, last_error_message FROM system.errors WHERE last_error_time >= toDateTime('" + start_time + "')" ) ) for error in errors: > assert (error in self.expect_errors) or (error in self.allow_errors) E AssertionError: assert ('NETLINK_ERROR' in ['QUERY_WAS_CANCELLED'] or 'NETLINK_ERROR' in []) E + where ['QUERY_WAS_CANCELLED'] = .expect_errors E + and [] = .allow_errors test_backup_restore_on_cluster/test_cancel_backup.py:394: AssertionError ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml ------------------------------ Captured log setup ------------------------------ 2026-04-10 19:11:45.214000 [ 658 ] DEBUG : Command:[docker ps | wc -l] (cluster.py:121, run_and_check) 2026-04-10 19:11:45.232000 [ 658 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 19:11:45.232000 [ 658 ] DEBUG : No running containers (conftest.py:95, cleanup_environment) 2026-04-10 19:11:45.232000 [ 658 ] DEBUG : Pruning Docker networks (conftest.py:97, cleanup_environment) 2026-04-10 19:11:45.232000 [ 658 ] DEBUG : Command:[docker network prune --force] (cluster.py:121, run_and_check) 2026-04-10 19:11:45.257000 [ 658 ] DEBUG : Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] (cluster.py:121, run_and_check) 2026-04-10 19:11:45.260000 [ 658 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:145, run_and_check) 2026-04-10 19:11:45.261000 [ 658 ] INFO : Running tests in /ClickHouse/tests/integration/test_backup_restore_on_cluster/test_cancel_backup.py (cluster.py:2748, start) 2026-04-10 19:11:45.261000 [ 658 ] DEBUG : Cluster start called. is_up=False (cluster.py:2755, start) 2026-04-10 19:11:45.280000 [ 658 ] DEBUG : Docker networks for project roottestbackuprestoreonclustercancelbackup-gw0 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 19:11:45.298000 [ 658 ] DEBUG : Docker containers for project roottestbackuprestoreonclustercancelbackup-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 19:11:45.316000 [ 658 ] DEBUG : Docker volumes for project roottestbackuprestoreonclustercancelbackup-gw0 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 19:11:45.317000 [ 658 ] DEBUG : Cleanup called (cluster.py:861, cleanup) 2026-04-10 19:11:45.335000 [ 658 ] DEBUG : Docker networks for project roottestbackuprestoreonclustercancelbackup-gw0 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 19:11:45.353000 [ 658 ] DEBUG : Docker containers for project roottestbackuprestoreonclustercancelbackup-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 19:11:45.376000 [ 658 ] DEBUG : Docker volumes for project roottestbackuprestoreonclustercancelbackup-gw0 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 19:11:45.377000 [ 658 ] DEBUG : Command:[docker container list --all --filter name='^/roottestbackuprestoreonclustercancelbackup-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2026-04-10 19:11:45.393000 [ 658 ] DEBUG : Unstopped containers: {} (cluster.py:875, cleanup) 2026-04-10 19:11:45.393000 [ 658 ] DEBUG : No running containers for project: roottestbackuprestoreonclustercancelbackup-gw0 (cluster.py:889, cleanup) 2026-04-10 19:11:45.393000 [ 658 ] DEBUG : Trying to prune unused networks... (cluster.py:895, cleanup) 2026-04-10 19:11:45.415000 [ 658 ] DEBUG : Trying to prune unused images... (cluster.py:911, cleanup) 2026-04-10 19:11:45.416000 [ 658 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2026-04-10 19:11:45.440000 [ 658 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2026-04-10 19:11:45.440000 [ 658 ] DEBUG : Images pruned (cluster.py:914, cleanup) 2026-04-10 19:11:45.441000 [ 658 ] DEBUG : Trying to prune unused volumes... (cluster.py:920, cleanup) 2026-04-10 19:11:45.441000 [ 658 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2026-04-10 19:11:45.462000 [ 658 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 19:11:45.462000 [ 658 ] DEBUG : Volumes pruned: 1 (cluster.py:925, cleanup) 2026-04-10 19:11:45.462000 [ 658 ] DEBUG : Setup directory for instance: node1 (cluster.py:2768, start) 2026-04-10 19:11:45.463000 [ 658 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4638, create_dir) 2026-04-10 19:11:45.463000 [ 658 ] DEBUG : Create directory for common tests configuration (cluster.py:4643, create_dir) 2026-04-10 19:11:45.463000 [ 658 ] DEBUG : Copy common configuration from helpers (cluster.py:4663, create_dir) 2026-04-10 19:11:45.464000 [ 658 ] DEBUG : Generate and write macros file (cluster.py:4715, create_dir) 2026-04-10 19:11:45.464000 [ 658 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/lesser_timeouts.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/slow_backups.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/shutdown_cancel_backups.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node1/configs/config.d (cluster.py:4755, create_dir) 2026-04-10 19:11:45.465000 [ 658 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node1/database (cluster.py:4772, create_dir) 2026-04-10 19:11:45.465000 [ 658 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node1/logs (cluster.py:4783, create_dir) 2026-04-10 19:11:45.465000 [ 658 ] DEBUG : Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" (cluster.py:4864, create_dir) 2026-04-10 19:11:45.465000 [ 658 ] INFO : external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/backups (cluster.py:4893, create_dir) 2026-04-10 19:11:45.466000 [ 658 ] DEBUG : Setup directory for instance: node2 (cluster.py:2768, start) 2026-04-10 19:11:45.466000 [ 658 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4638, create_dir) 2026-04-10 19:11:45.466000 [ 658 ] DEBUG : Create directory for common tests configuration (cluster.py:4643, create_dir) 2026-04-10 19:11:45.466000 [ 658 ] DEBUG : Copy common configuration from helpers (cluster.py:4663, create_dir) 2026-04-10 19:11:45.467000 [ 658 ] DEBUG : Generate and write macros file (cluster.py:4715, create_dir) 2026-04-10 19:11:45.467000 [ 658 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/lesser_timeouts.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/slow_backups.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/shutdown_cancel_backups.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node2/configs/config.d (cluster.py:4755, create_dir) 2026-04-10 19:11:45.468000 [ 658 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node2/database (cluster.py:4772, create_dir) 2026-04-10 19:11:45.468000 [ 658 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node2/logs (cluster.py:4783, create_dir) 2026-04-10 19:11:45.468000 [ 658 ] DEBUG : Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" (cluster.py:4864, create_dir) 2026-04-10 19:11:45.468000 [ 658 ] INFO : external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/backups (cluster.py:4893, create_dir) 2026-04-10 19:11:45.468000 [ 658 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'LSAN_OPTIONS': 'suppressions=/etc/clickhouse-server/lsan_suppressions.txt', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:5ccda723c1fc', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/.env (cluster.py:96, _create_env_file) 2026-04-10 19:11:45.469000 [ 658 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2026-04-10 19:11:45.469000 [ 658 ] DEBUG : No config file found (config.py:28, find_config_file) 2026-04-10 19:11:45.469000 [ 658 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2026-04-10 19:11:45.469000 [ 658 ] DEBUG : No config file found (config.py:28, find_config_file) 2026-04-10 19:11:45.480000 [ 658 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2026-04-10 19:11:45.481000 [ 658 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/.env --project-name roottestbackuprestoreonclustercancelbackup-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node2/docker-compose.yml pull] (cluster.py:121, run_and_check) 2026-04-10 19:11:55.933000 [ 658 ] DEBUG : Stderr: node1 Skipped - Image is already being pulled by node2 (cluster.py:147, run_and_check) 2026-04-10 19:11:55.933000 [ 658 ] DEBUG : Stderr: zoo1 Skipped - Image is already being pulled by node2 (cluster.py:147, run_and_check) 2026-04-10 19:11:55.934000 [ 658 ] DEBUG : Stderr: zoo2 Skipped - Image is already being pulled by node2 (cluster.py:147, run_and_check) 2026-04-10 19:11:55.934000 [ 658 ] DEBUG : Stderr: zoo3 Skipped - Image is already being pulled by node2 (cluster.py:147, run_and_check) 2026-04-10 19:11:55.934000 [ 658 ] DEBUG : Stderr: node2 Pulling (cluster.py:147, run_and_check) 2026-04-10 19:11:55.934000 [ 658 ] DEBUG : Stderr: node2 Pulled (cluster.py:147, run_and_check) 2026-04-10 19:11:55.934000 [ 658 ] DEBUG : Setup ZooKeeper (cluster.py:2809, start) 2026-04-10 19:11:55.934000 [ 658 ] DEBUG : Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper1/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper1/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper1/coordination', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper2/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper2/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper2/coordination', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper3/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper3/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/keeper3/coordination'] (cluster.py:2810, start) 2026-04-10 19:11:55.935000 [ 658 ] DEBUG : Command:[docker compose --project-name roottestbackuprestoreonclustercancelbackup-gw0 --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] (cluster.py:121, run_and_check) 2026-04-10 19:11:56.484000 [ 658 ] DEBUG : Stderr:time="2026-04-10T19:11:55Z" level=trace msg="Docker Desktop integration not enabled" (cluster.py:147, run_and_check) 2026-04-10 19:11:56.484000 [ 658 ] DEBUG : Stderr: Network roottestbackuprestoreonclustercancelbackup-gw0_default Creating (cluster.py:147, run_and_check) 2026-04-10 19:11:56.484000 [ 658 ] DEBUG : Stderr: Network roottestbackuprestoreonclustercancelbackup-gw0_default Created (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo3-1 Creating (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo1-1 Creating (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo2-1 Creating (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo1-1 Created (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo2-1 Created (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo3-1 Created (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo2-1 Starting (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo3-1 Starting (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo1-1 Starting (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo2-1 Started (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo1-1 Started (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo3-1 Started (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr:time="2026-04-10T19:11:56Z" level=debug msg="otel error" error="" (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Stderr:time="2026-04-10T19:11:56Z" level=debug msg="otel error" error="" (cluster.py:147, run_and_check) 2026-04-10 19:11:56.485000 [ 658 ] DEBUG : Wait ZooKeeper to start (cluster.py:2446, wait_zookeeper_to_start) 2026-04-10 19:11:56.486000 [ 658 ] DEBUG : get_instance_ip instance_name=zoo1 (cluster.py:2015, get_instance_ip) 2026-04-10 19:11:56.488000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclustercancelbackup-gw0-zoo1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:11:56.488000 [ 658 ] DEBUG : get_kazoo_client: zoo1, ip:172.16.1.4, port:2181, use_ssl:False (cluster.py:3322, get_kazoo_client) 2026-04-10 19:11:56.489000 [ 658 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:11:56.490000 [ 658 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 19:11:56.630000 [ 658 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:11:56.630000 [ 658 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 19:11:56.825000 [ 658 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:11:56.826000 [ 658 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 19:11:57.327000 [ 658 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:11:57.328000 [ 658 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 19:11:58.011000 [ 658 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:11:58.012000 [ 658 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 19:11:59.679000 [ 658 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:11:59.680000 [ 658 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2026-04-10 19:11:59.685000 [ 658 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2026-04-10 19:11:59.686000 [ 658 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2026-04-10 19:11:59.687000 [ 658 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2026-04-10 19:11:59.688000 [ 658 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2026-04-10 19:11:59.690000 [ 658 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2026-04-10 19:11:59.691000 [ 658 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2026-04-10 19:11:59.691000 [ 658 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2026-04-10 19:11:59.791000 [ 658 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2026-04-10 19:11:59.792000 [ 658 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2026-04-10 19:11:59.792000 [ 658 ] DEBUG : get_instance_ip instance_name=zoo2 (cluster.py:2015, get_instance_ip) 2026-04-10 19:11:59.794000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclustercancelbackup-gw0-zoo2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:11:59.795000 [ 658 ] DEBUG : get_kazoo_client: zoo2, ip:172.16.1.3, port:2181, use_ssl:False (cluster.py:3322, get_kazoo_client) 2026-04-10 19:11:59.796000 [ 658 ] INFO : Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:11:59.796000 [ 658 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2026-04-10 19:11:59.799000 [ 658 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2026-04-10 19:11:59.800000 [ 658 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2026-04-10 19:11:59.801000 [ 658 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2026-04-10 19:11:59.801000 [ 658 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2026-04-10 19:11:59.804000 [ 658 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2026-04-10 19:11:59.804000 [ 658 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2026-04-10 19:11:59.804000 [ 658 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2026-04-10 19:11:59.904000 [ 658 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2026-04-10 19:11:59.905000 [ 658 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2026-04-10 19:11:59.905000 [ 658 ] DEBUG : get_instance_ip instance_name=zoo3 (cluster.py:2015, get_instance_ip) 2026-04-10 19:11:59.908000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclustercancelbackup-gw0-zoo3-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:11:59.909000 [ 658 ] DEBUG : get_kazoo_client: zoo3, ip:172.16.1.2, port:2181, use_ssl:False (cluster.py:3322, get_kazoo_client) 2026-04-10 19:11:59.911000 [ 658 ] INFO : Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:11:59.911000 [ 658 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2026-04-10 19:11:59.914000 [ 658 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2026-04-10 19:11:59.914000 [ 658 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2026-04-10 19:11:59.915000 [ 658 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2026-04-10 19:11:59.916000 [ 658 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2026-04-10 19:11:59.918000 [ 658 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2026-04-10 19:11:59.918000 [ 658 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2026-04-10 19:11:59.919000 [ 658 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2026-04-10 19:12:00.019000 [ 658 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2026-04-10 19:12:00.020000 [ 658 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2026-04-10 19:12:00.020000 [ 658 ] DEBUG : All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') (cluster.py:2462, wait_zookeeper_nodes_to_start) 2026-04-10 19:12:00.021000 [ 658 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/.env --project-name roottestbackuprestoreonclustercancelbackup-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node2/docker-compose.yml up -d --no-recreate') (cluster.py:3149, start) 2026-04-10 19:12:00.021000 [ 658 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/.env --project-name roottestbackuprestoreonclustercancelbackup-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node2/docker-compose.yml up -d --no-recreate] (cluster.py:121, run_and_check) 2026-04-10 19:12:00.481000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo3-1 Running (cluster.py:147, run_and_check) 2026-04-10 19:12:00.481000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo1-1 Running (cluster.py:147, run_and_check) 2026-04-10 19:12:00.481000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo2-1 Running (cluster.py:147, run_and_check) 2026-04-10 19:12:00.481000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node1-1 Creating (cluster.py:147, run_and_check) 2026-04-10 19:12:00.481000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node2-1 Creating (cluster.py:147, run_and_check) 2026-04-10 19:12:00.481000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node1-1 Created (cluster.py:147, run_and_check) 2026-04-10 19:12:00.482000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node2-1 Created (cluster.py:147, run_and_check) 2026-04-10 19:12:00.482000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node1-1 Starting (cluster.py:147, run_and_check) 2026-04-10 19:12:00.482000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node2-1 Starting (cluster.py:147, run_and_check) 2026-04-10 19:12:00.482000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node1-1 Started (cluster.py:147, run_and_check) 2026-04-10 19:12:00.482000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node2-1 Started (cluster.py:147, run_and_check) 2026-04-10 19:12:00.482000 [ 658 ] DEBUG : ClickHouse instance created (cluster.py:3157, start) 2026-04-10 19:12:00.482000 [ 658 ] DEBUG : get_instance_ip instance_name=node1 (cluster.py:2015, get_instance_ip) 2026-04-10 19:12:00.484000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclustercancelbackup-gw0-node1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:00.485000 [ 658 ] DEBUG : get_instance_ip instance_name=node1 (cluster.py:2025, get_instance_global_ipv6) 2026-04-10 19:12:00.487000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclustercancelbackup-gw0-node1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:00.488000 [ 658 ] DEBUG : Waiting for ClickHouse start in node1, ip: 172.16.1.6... (cluster.py:3165, start) 2026-04-10 19:12:00.491000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclustercancelbackup-gw0-node1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:00.494000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/acaef689df5bc8ee1f394234c8b12ba29ca792640581a2e470c35d38d174601d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:00.597000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/acaef689df5bc8ee1f394234c8b12ba29ca792640581a2e470c35d38d174601d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:00.700000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/acaef689df5bc8ee1f394234c8b12ba29ca792640581a2e470c35d38d174601d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:00.803000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/acaef689df5bc8ee1f394234c8b12ba29ca792640581a2e470c35d38d174601d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:00.906000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/acaef689df5bc8ee1f394234c8b12ba29ca792640581a2e470c35d38d174601d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:01.009000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/acaef689df5bc8ee1f394234c8b12ba29ca792640581a2e470c35d38d174601d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:01.009000 [ 658 ] DEBUG : ClickHouse node1 started (cluster.py:3169, start) 2026-04-10 19:12:01.010000 [ 658 ] DEBUG : get_instance_ip instance_name=node2 (cluster.py:2015, get_instance_ip) 2026-04-10 19:12:01.012000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclustercancelbackup-gw0-node2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:01.012000 [ 658 ] DEBUG : get_instance_ip instance_name=node2 (cluster.py:2025, get_instance_global_ipv6) 2026-04-10 19:12:01.013000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclustercancelbackup-gw0-node2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:01.014000 [ 658 ] DEBUG : Waiting for ClickHouse start in node2, ip: 172.16.1.5... (cluster.py:3165, start) 2026-04-10 19:12:01.015000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclustercancelbackup-gw0-node2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:01.017000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/containers/7de0ca803761bf9f2e1570dcd9a63952f7e5a4d2aa85193405dfdbc95c2c95ed/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:12:01.018000 [ 658 ] DEBUG : ClickHouse node2 started (cluster.py:3169, start) ----------------------------- Captured stdout call ----------------------------- Using node1 as initiator Waiting for number of system processes = 1+ Got 1 system processes for backup e7f565f649124a3a94c119e813c5e56a after waiting 1 seconds node2: Restarting... node2: Restarted Waiting for number of system processes = 0 Got 0 system processes for backup e7f565f649124a3a94c119e813c5e56a after waiting 0 seconds node1: Found errors: ['QUERY_WAS_CANCELLED'] QUERY_WAS_CANCELLED Got error from host node2:9000. DB::Exception: Query was cancelled. Stack trace:\n\n0. DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000d7ed77b\n1. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000008292ccc\n2. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x000000000829f48b\n3. DB::QueryStatus::throwQueryWasCancelled() const @ 0x00000000119f1b2f\n4. DB::QueryStatus::throwProperExceptionIfNeeded(unsigned long const&, unsigned long const&) @ 0x00000000119f1a2c\n5. DB::BackupEntriesCollector::setStage(String const&, String const&) @ 0x00000000107a103f\n6. DB::BackupEntriesCollector::gatherMetadataAndCheckConsistency() @ 0x000000001079db85\n7. DB::BackupEntriesCollector::run() @ 0x000000001079a326\n8. DB::BackupsWorker::BackupStarter::doBackup() @ 0x00000000107d9a61\n9. void std::__function::__policy_invoker::__call_impl[abi:ne190107] const&, std::shared_ptr const&)::$_0, void ()>>(std::__function::__policy_storage const*) @ 0x00000000107d77ac\n10. std::__packaged_task_func (std::function&&, Priority)> DB::threadPoolCallbackRunnerUnsafe>(ThreadPoolImpl>&, String const&)::\'lambda\'(std::function&&, Priority)::operator()(std::function&&, Priority)::\'lambda\'(), std::allocator (std::function&&, Priority)> DB::threadPoolCallbackRunnerUnsafe>(ThreadPoolImpl>&, String const&)::\'lambda\'(std::function&&, Priority)::operator()(std::function&&, Priority)::\'lambda\'()>, void ()>::operator()() @ 0x0000000010290f04\n11. std::packaged_task::operator()() @ 0x000000001029106c\n12. ThreadPoolImpl>::ThreadFromThreadPool::worker() @ 0x000000000d92283b\n13. void std::__function::__policy_invoker::__call_impl[abi:ne190107]::ThreadFromGlobalPoolImpl>::ThreadFromThreadPool::*)(), ThreadPoolImpl>::ThreadFromThreadPool*>(void (ThreadPoolImpl>::ThreadFromThreadPool::*&&)(), ThreadPoolImpl>::ThreadFromThreadPool*&&)::\'lambda\'(), void ()>>(std::__function::__policy_storage const*) @ 0x000000000d929363\n14. ThreadPoolImpl::ThreadFromThreadPool::worker() @ 0x000000000d91fb6f\n15. void* std::__thread_proxy[abi:ne190107]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*>>(void*) @ 0x000000000d926e3a\n16. ? @ 0x00007fd89f8f6ac3\n17. ? @ 0x00007fd89f988850\n\nJob\'s origin stack trace:\n0. void boost::heap::priority_queue, boost::parameter::void_, boost::parameter::void_, boost::parameter::void_>::emplace, Priority&, StrongTypedef&, DB::OpenTelemetry::TracingContextOnThread const, bool&, (anonymous namespace)::ScopedDecrement>(std::function&&, Priority&, StrongTypedef&, DB::OpenTelemetry::TracingContextOnThread const&&, bool&, (anonymous namespace)::ScopedDecrement&&) @ 0x000000000d927a77\n1. void ThreadPoolImpl>::scheduleImpl(std::function, Priority, std::optional, bool) @ 0x000000000d923c18\n2. std::function (std::function&&, Priority)> DB::threadPoolCallbackRunnerUnsafe>(ThreadPoolImpl>&, String const&)::\'lambda\'(std::function&&, Priority)::operator()(std::function&&, Priority) @ 0x000000001029071a\n3. std::future std::__function::__policy_invoker (std::function&&, Priority)>::__call_impl[abi:ne190107] (std::function&&, Priority)> DB::threadPoolCallbackRunnerUnsafe>(ThreadPoolImpl>&, String const&)::\'lambda\'(std::function&&, Priority), std::future (std::function&&, Priority)>>(std::__function::__policy_storage const*, std::function&&, Priority&&) @ 0x000000001029054f\n4. DB::BackupsWorker::startMakingBackup(std::shared_ptr const&, std::shared_ptr const&) @ 0x00000000107ce5d5\n5. DB::BackupsWorker::start(std::shared_ptr const&, std::shared_ptr) @ 0x00000000107cd5e3\n6. DB::InterpreterBackupQuery::execute() @ 0x0000000011d3a5a4\n7. DB::executeQueryImpl(char const*, char const*, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum, DB::ReadBuffer*, std::shared_ptr&) @ 0x0000000011cadfab\n8. DB::executeQuery(DB::ReadBuffer&, DB::WriteBuffer&, bool, std::shared_ptr, std::function, DB::QueryFlags, std::optional const&, std::function const&, std::optional const&)>) @ 0x0000000011cb22a3\n9. DB::DDLWorker::tryExecuteQuery(DB::DDLTaskBase&, std::shared_ptr const&, bool) @ 0x00000000114c2765\n10. DB::DDLWorker::processTask(DB::DDLTaskBase&, std::shared_ptr const&, bool) @ 0x00000000114c0afb\n11. DB::DDLWorker::scheduleTasks(bool) @ 0x00000000114bd722\n12. DB::DDLWorker::runMainThread() @ 0x00000000114b562c\n13. void std::__function::__policy_invoker::__call_impl[abi:ne190107]::ThreadFromGlobalPoolImpl(void (DB::DDLWorker::*&&)(), DB::DDLWorker*&&)::\'lambda\'(), void ()>>(std::__function::__policy_storage const*) @ 0x00000000114d4543\n14. ThreadPoolImpl::ThreadFromThreadPool::worker() @ 0x000000000d91fb6f\n15. void* std::__thread_proxy[abi:ne190107]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*>>(void*) @ 0x000000000d926e3a\n16. ? @ 0x00007fd89f8f6ac3\n17. ? @ 0x00007fd89f988850\n node2: Found errors: ['NETLINK_ERROR'] NO_ELEMENTS_IN_CONFIG Certificate file is not set. NETLINK_ERROR Can\'t receive Netlink response: error -2 ------------------------------ Captured log call ------------------------------- 2026-04-10 19:12:02.021000 [ 658 ] DEBUG : Executing query SELECT name FROM system.zookeeper WHERE path = '/clickhouse/backups' AND NOT (name == 'alive_tracker') on node1 (cluster.py:3658, query) 2026-04-10 19:12:02.137000 [ 658 ] DEBUG : Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}') ORDER BY tuple() PARTITION BY x%10 on node2 (cluster.py:3658, query) 2026-04-10 19:12:02.302000 [ 658 ] DEBUG : Executing query INSERT INTO tbl SELECT number FROM numbers(10) on node2 (cluster.py:3658, query) 2026-04-10 19:12:02.468000 [ 658 ] DEBUG : Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', 'e7f565f649124a3a94c119e813c5e56a') SETTINGS id='e7f565f649124a3a94c119e813c5e56a' ASYNC on node1 (cluster.py:3658, query) 2026-04-10 19:12:02.533000 [ 658 ] DEBUG : Executing query SELECT status FROM system.backups WHERE id='e7f565f649124a3a94c119e813c5e56a' on node1 (cluster.py:3658, query) 2026-04-10 19:12:02.598000 [ 658 ] DEBUG : Executing query SELECT count() FROM system.processes WHERE (query_kind='Backup') AND (query LIKE '%e7f565f649124a3a94c119e813c5e56a%') on node1 (cluster.py:3658, query) 2026-04-10 19:12:02.713000 [ 658 ] DEBUG : Executing query SELECT count() FROM system.processes WHERE (query_kind='Backup') AND (query LIKE '%e7f565f649124a3a94c119e813c5e56a%') on node2 (cluster.py:3658, query) 2026-04-10 19:12:03.779000 [ 658 ] DEBUG : Executing query SELECT count() FROM system.processes WHERE (query_kind='Backup') AND (query LIKE '%e7f565f649124a3a94c119e813c5e56a%') on node2 (cluster.py:3658, query) 2026-04-10 19:12:03.894000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:03.895000 [ 658 ] DEBUG : Command:[docker exec -u root roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps -C clickhouse] (cluster.py:121, run_and_check) 2026-04-10 19:12:03.955000 [ 658 ] DEBUG : Stdout: PID TTY TIME CMD (cluster.py:145, run_and_check) 2026-04-10 19:12:03.955000 [ 658 ] DEBUG : Stdout: 10 ? 00:00:00 clickhouse (cluster.py:145, run_and_check) 2026-04-10 19:12:03.956000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:03.956000 [ 658 ] DEBUG : Command:[docker exec -u root roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c pkill clickhouse] (cluster.py:121, run_and_check) 2026-04-10 19:12:04.003000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:04.003000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:04.045000 [ 658 ] DEBUG : Stdout:10 (cluster.py:145, run_and_check) 2026-04-10 19:12:05.047000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:05.047000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:05.093000 [ 658 ] DEBUG : Stdout:10 (cluster.py:145, run_and_check) 2026-04-10 19:12:06.094000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:06.095000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:06.158000 [ 658 ] DEBUG : Stdout:10 (cluster.py:145, run_and_check) 2026-04-10 19:12:07.160000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:07.161000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:07.220000 [ 658 ] DEBUG : Stdout:10 (cluster.py:145, run_and_check) 2026-04-10 19:12:08.221000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:08.222000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:08.284000 [ 658 ] DEBUG : Stdout:10 (cluster.py:145, run_and_check) 2026-04-10 19:12:09.286000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:09.286000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:09.336000 [ 658 ] DEBUG : Stdout:10 (cluster.py:145, run_and_check) 2026-04-10 19:12:10.338000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:10.338000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:10.380000 [ 658 ] DEBUG : Stdout:10 (cluster.py:145, run_and_check) 2026-04-10 19:12:11.381000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:11.382000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:11.428000 [ 658 ] DEBUG : Stdout:10 (cluster.py:145, run_and_check) 2026-04-10 19:12:12.429000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:12.430000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:12.482000 [ 658 ] DEBUG : Stdout:10 (cluster.py:145, run_and_check) 2026-04-10 19:12:13.483000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:13.483000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:13.527000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:13.528000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:13.580000 [ 658 ] DEBUG : No clickhouse process running. Start new one. (cluster.py:4024, start_clickhouse) 2026-04-10 19:12:13.584000 [ 658 ] DEBUG : http://localhost:None "POST /v1.46/containers/roottestbackuprestoreonclustercancelbackup-gw0-node2-1/exec HTTP/1.1" 201 74 (connectionpool.py:547, _make_request) 2026-04-10 19:12:13.618000 [ 658 ] DEBUG : http://localhost:None "POST /v1.46/exec/60d8f814b87c79f0f283727bb2ceb34019dceb16c14fd13dc9b60a60f5a96d7e/start HTTP/1.1" 200 0 (connectionpool.py:547, _make_request) 2026-04-10 19:12:13.620000 [ 658 ] DEBUG : http://localhost:None "GET /v1.46/exec/60d8f814b87c79f0f283727bb2ceb34019dceb16c14fd13dc9b60a60f5a96d7e/json HTTP/1.1" 200 585 (connectionpool.py:547, _make_request) 2026-04-10 19:12:14.622000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:14.622000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:14.684000 [ 658 ] DEBUG : Stdout:841 (cluster.py:145, run_and_check) 2026-04-10 19:12:14.684000 [ 658 ] DEBUG : Clickhouse process running. (cluster.py:4038, start_clickhouse) 2026-04-10 19:12:14.685000 [ 658 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2061, exec_in_container) 2026-04-10 19:12:14.685000 [ 658 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:14.747000 [ 658 ] DEBUG : Stdout:841 (cluster.py:145, run_and_check) 2026-04-10 19:12:14.748000 [ 658 ] DEBUG : Executing query select 20 on node2 (cluster.py:3658, query) 2026-04-10 19:12:14.813000 [ 658 ] DEBUG : Executing query SELECT count() FROM system.processes WHERE (query_kind='Backup') AND (query LIKE '%e7f565f649124a3a94c119e813c5e56a%') on node1 (cluster.py:3658, query) 2026-04-10 19:12:14.928000 [ 658 ] DEBUG : Executing query SELECT count() FROM system.processes WHERE (query_kind='Backup') AND (query LIKE '%e7f565f649124a3a94c119e813c5e56a%') on node2 (cluster.py:3658, query) 2026-04-10 19:12:15.044000 [ 658 ] DEBUG : Executing query SELECT status FROM system.backups WHERE id='e7f565f649124a3a94c119e813c5e56a' on node1 (cluster.py:3658, query) 2026-04-10 19:12:15.109000 [ 658 ] DEBUG : Executing query SELECT error FROM system.backups WHERE id='e7f565f649124a3a94c119e813c5e56a' on node1 (cluster.py:3658, query) 2026-04-10 19:12:15.174000 [ 658 ] DEBUG : Executing query SYSTEM FLUSH LOGS on node1 (cluster.py:3658, query) 2026-04-10 19:12:15.340000 [ 658 ] DEBUG : Executing query SELECT status FROM system.backup_log WHERE id='e7f565f649124a3a94c119e813c5e56a' ORDER BY status on node1 (cluster.py:3658, query) 2026-04-10 19:12:15.405000 [ 658 ] DEBUG : Executing query SELECT name FROM system.zookeeper WHERE path = '/clickhouse/backups' AND NOT (name == 'alive_tracker') on node1 (cluster.py:3658, query) 2026-04-10 19:12:15.470000 [ 658 ] DEBUG : Executing query SELECT name FROM system.errors WHERE last_error_time >= toDateTime('2026-04-10 19:12:02') AND NOT ((name == 'KEEPER_EXCEPTION') AND (last_error_message LIKE '%Fault injection%')) AND NOT (name == 'NO_ELEMENTS_IN_CONFIG') on node1 (cluster.py:3658, query) 2026-04-10 19:12:15.536000 [ 658 ] DEBUG : Executing query SELECT name, last_error_message FROM system.errors WHERE last_error_time >= toDateTime('2026-04-10 19:12:02') on node1 (cluster.py:3658, query) 2026-04-10 19:12:15.601000 [ 658 ] DEBUG : Executing query SELECT name FROM system.errors WHERE last_error_time >= toDateTime('2026-04-10 19:12:02') AND NOT ((name == 'KEEPER_EXCEPTION') AND (last_error_message LIKE '%Fault injection%')) AND NOT (name == 'NO_ELEMENTS_IN_CONFIG') on node2 (cluster.py:3658, query) 2026-04-10 19:12:15.666000 [ 658 ] DEBUG : Executing query SELECT name, last_error_message FROM system.errors WHERE last_error_time >= toDateTime('2026-04-10 19:12:02') on node2 (cluster.py:3658, query) ---------------------------- Captured log teardown ----------------------------- 2026-04-10 19:12:15.802000 [ 658 ] DEBUG : Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster' SYNC on node1 (cluster.py:3658, query) 2026-04-10 19:12:15.917000 [ 658 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/.env --project-name roottestbackuprestoreonclustercancelbackup-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node2/docker-compose.yml stop --timeout 20] (cluster.py:121, run_and_check) 2026-04-10 19:12:16.699000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:12:16.699000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node2-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:12:16.699000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node2-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:12:16.700000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:12:16.700000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo3-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:12:16.700000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:12:16.700000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo2-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:12:16.700000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo3-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:12:16.700000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:12:16.700000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo2-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:12:16.700000 [ 658 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:121, run_and_check) 2026-04-10 19:12:16.714000 [ 658 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:121, run_and_check) 2026-04-10 19:12:16.727000 [ 658 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/.env --project-name roottestbackuprestoreonclustercancelbackup-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-1-gw0/node2/docker-compose.yml down --volumes] (cluster.py:121, run_and_check) 2026-04-10 19:12:17.176000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:12:17.176000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node2-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node1-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node2-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node2-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node1-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-node2-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo3-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo2-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo3-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo3-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo1-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo2-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:12:17.177000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo2-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:12:17.178000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo3-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:12:17.178000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo1-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:12:17.178000 [ 658 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw0-zoo2-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:12:17.178000 [ 658 ] DEBUG : Stderr: Network roottestbackuprestoreonclustercancelbackup-gw0_default Removing (cluster.py:147, run_and_check) 2026-04-10 19:12:17.178000 [ 658 ] DEBUG : Stderr: Network roottestbackuprestoreonclustercancelbackup-gw0_default Removed (cluster.py:147, run_and_check) 2026-04-10 19:12:17.178000 [ 658 ] DEBUG : Cleanup called (cluster.py:861, cleanup) 2026-04-10 19:12:17.205000 [ 658 ] DEBUG : Docker networks for project roottestbackuprestoreonclustercancelbackup-gw0 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 19:12:17.223000 [ 658 ] DEBUG : Docker containers for project roottestbackuprestoreonclustercancelbackup-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 19:12:17.245000 [ 658 ] DEBUG : Docker volumes for project roottestbackuprestoreonclustercancelbackup-gw0 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 19:12:17.245000 [ 658 ] DEBUG : Command:[docker container list --all --filter name='^/roottestbackuprestoreonclustercancelbackup-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2026-04-10 19:12:17.262000 [ 658 ] DEBUG : Unstopped containers: {} (cluster.py:875, cleanup) 2026-04-10 19:12:17.262000 [ 658 ] DEBUG : No running containers for project: roottestbackuprestoreonclustercancelbackup-gw0 (cluster.py:889, cleanup) 2026-04-10 19:12:17.262000 [ 658 ] DEBUG : Trying to prune unused networks... (cluster.py:895, cleanup) 2026-04-10 19:12:17.280000 [ 658 ] DEBUG : Trying to prune unused images... (cluster.py:911, cleanup) 2026-04-10 19:12:17.280000 [ 658 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2026-04-10 19:12:17.306000 [ 658 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2026-04-10 19:12:17.306000 [ 658 ] DEBUG : Images pruned (cluster.py:914, cleanup) 2026-04-10 19:12:17.306000 [ 658 ] DEBUG : Trying to prune unused volumes... (cluster.py:920, cleanup) 2026-04-10 19:12:17.306000 [ 658 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2026-04-10 19:12:17.329000 [ 658 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 19:12:17.329000 [ 658 ] DEBUG : Volumes pruned: 1 (cluster.py:925, cleanup) ----------------- generated report log file: parallel0_1.jsonl ----------------- ============================== slowest durations =============================== 15.81s setup test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup 14.71s call test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup 1.53s teardown test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup =========================== short test summary info ============================ FAILED test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup ============================== 1 failed in 44.03s ============================== Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 492, in subprocess.check_call(cmd, shell=True, bufsize=0) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_vlduwg --privileged --dns-search='.' --memory=30709047296 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=5ccda723c1fc -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=d862517635bf -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=1 --color=no --durations=0 --report-log=parallel0_1.jsonl --report-log-exclude-logs-on-passed-tests test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup -vvv " altinityinfra/integration-tests-runner:226bfaf75ac1 ' returned non-zero exit status 1.